-
-
Notifications
You must be signed in to change notification settings - Fork 3k
Update mistral-medium prices and context sizes #10729
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
While testing the Mistral model, I noticed a discrepancy in the pricing shown on the logs screen. After reviewing the code, I confirmed that the pricing values were incorrect. This PR corrects the input and output token pricing for the latest Mistral model and adds the newly released mistral-medium-2505 version.
The latest updates on your projects. Learn more about Vercel for Git ↗︎
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The context sizes are incorrect - The previous values were correct.
@emerzon Mistral medium latest is only 32k?! Every place that I checked shows 128k. |
Does mistral/mistral-medium points to the old one? |
@regismesquita It's probably better to add a new entry for the new model matching the API naming from mistral: Also, mind that the the output context size ( |
This very link that you shared shows max tokens at 128k. I know they might be actually talking about the input context there... but going by the word... I have also found sources for output at 128k but none reliable , only stuff like openrouter listing of the new model.
I will remove mistral-medium name from the scope of this PR and retain only the 2503 and the latest identifiers.
I will try to experiment with this model today and also try to contact mistral.
|
sets mistral medium alias to the old values as it probably points to the old version.
@emerzon I have reverted mistral/mistral-medium to the old values, and now only the |
128k is for input tokens - not output tokens. |
@emerzon can you share how are you sure of that? I couldn't confirm that anywhere in their docs. |
@emerzon haven't got any response yet, and I do think that it is less than 128k but since I haven't got any response from them I prefer to err on the safe side. so I changed it down to 8191 , and I can follow up later if I found any evidence against it. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Ignore this comment. |
While testing the Mistral model, I noticed a discrepancy in the pricing shown on the logs screen. After reviewing the code, I confirmed that the pricing values were incorrect.
This PR corrects the input and output token pricing for the latest Mistral model and adds the newly released mistral-medium-2505 version.
Title
Relevant issues
Pre-Submission checklist
Please complete all items before asking a LiteLLM maintainer to review your PR
tests/litellm/
directory, Adding at least 1 test is a hard requirement - see detailsmake test-unit
Type
🆕 New Feature
🐛 Bug Fix
🧹 Refactoring
📖 Documentation
🚄 Infrastructure
✅ Test
Changes